Goto

Collaborating Authors

 latent dimension




e04101138a3c94544760c1dbdf2c7a2d-Paper-Conference.pdf

Neural Information Processing Systems

For example, while prior work has suggested that theglobally optimal VAEsolution canlearn thecorrect manifold dimension, anecessary (butnotsufficient)condition forproducing samplesfrom the true data distribution, this has never been rigorously proven. Moreover, it remains unclear how such considerations would change when various types of conditioning variablesare introduced, or when the data support is extended to a union of manifolds (e.g., as is likely the case for MNIST digits and related). In this work, we address these points by first proving that VAE global minima are indeed capable of recovering the correct manifold dimension.




bfd2308e9e75263970f8079115edebbd-Supplemental.pdf

Neural Information Processing Systems

We note that we modified the reference implementation provided with that work from in order to have consistent 4 4 kernels as shown in the architecture in Table 1.



87213955efbe48b46586e37bf2f1fe5b-Paper-Conference.pdf

Neural Information Processing Systems

Autoencoders (AEs) [1, 2] and its modern variants like the widely used variational autoencoders (VAEs) [3], are a powerful paradigm for self-supervised representation learning for generative modeling [4], compression [5], anomaly detection [6] or natural language processing [7]. Since autoencoders can learn low dimensional representations without requiring labeled data, they are particularly useful forcomputer vision taskswhere samples canbeveryhighdimensional making processing, transmitting, and search prohibitively expensive.